Goto

Collaborating Authors

 bootstrapping upper confidence bound


Reviews: Bootstrapping Upper Confidence Bound

Neural Information Processing Systems

I should be acknowledged that it is significantly more complex that UCB1 for example. Indeed at each time step B bootstrap repetitions are needed to estimated the bootstrapped quantiles, and each of them require to drawn n_k random variables for each arm k (the values of w's). Also, this requires to store the past rewards obtained on all arms, which requires a lot a memory. This constraint is also needed for the empirical KL-UCB mentioned above, which is one more reason to compare the two algorithms that have similar complexity. From Theorem 2, I guess that the w's are Rademacher random variables, but it would be good to specify this in the statement of the algorithm.


Reviews: Bootstrapping Upper Confidence Bound

Neural Information Processing Systems

The reviewers updated their scores after the rebuttal and discussion. Congratulations on a nice paper that had a consensus on acceptance! The reviewers has a couple of outstanding concerns (like relating B,T) that I would like the authors to explicitly discuss (including potentially mentioning open problems) in the camera-ready version.


Bootstrapping Upper Confidence Bound

Neural Information Processing Systems

Upper Confidence Bound (UCB) method is arguably the most celebrated one used in online decision making with partial information feedback. Existing techniques for constructing confidence bounds are typically built upon various concentration inequalities, which thus lead to over-exploration. In this paper, we propose a non-parametric and data-dependent UCB algorithm based on the multiplier bootstrap. To improve its finite sample performance, we further incorporate second-order correction into the above construction. In theory, we derive both problem-dependent and problem-independent regret bounds for multi-armed bandits under a much weaker tail assumption than the standard sub-Gaussianity.


Bootstrapping Upper Confidence Bound

Hao, Botao, Yadkori, Yasin Abbasi, Wen, Zheng, Cheng, Guang

Neural Information Processing Systems

Upper Confidence Bound (UCB) method is arguably the most celebrated one used in online decision making with partial information feedback. Existing techniques for constructing confidence bounds are typically built upon various concentration inequalities, which thus lead to over-exploration. In this paper, we propose a non-parametric and data-dependent UCB algorithm based on the multiplier bootstrap. To improve its finite sample performance, we further incorporate second-order correction into the above construction. In theory, we derive both problem-dependent and problem-independent regret bounds for multi-armed bandits under a much weaker tail assumption than the standard sub-Gaussianity.